Improving Naive Bayesian Classifier by Discriminative Training

نویسندگان

  • Kaizhu Huang
  • Zhangbing Zhou
  • Hai Dian Nan Lu
  • Irwin King
  • Michael R. Lyu
چکیده

Discriminative classifiers such as Support Vector Machines (SVM) directly learn a discriminant function or a posterior probability model to perform classification. On the other hand, generative classifiers often learn a joint probability model and then use the Bayes rule to construct a posterior classifier. In general, generative classifiers are not as accurate as discriminative classifiers. However generative classifiers provide a principled way to deal with the missing information problem, which discriminative classifiers cannot easily handle. To achieve good performance in various classification tasks, it is better to combine these two strategies. In this paper, we develop a method to train one of the popular generative classifiers, the Naive Bayesian classifier (NB) in a discriminative way. We name this new model as the Discriminative Naive Bayesian classifier. We provide theoretic justifications, outline the algorithm, and perform a serious of experiments on benchmark real-world datasets to demonstrate our model’s advantages. Its performance outperforms NB in classification tasks and outperforms SVM in handling missing information tasks.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Hybrid Generative/Discriminative Bayesian Classifier

In this paper, we introduce a new restricted Bayesian network classifier that extends naive Bayes by relaxing the conditional independence assumptions, and show that it is partly generative and partly discriminative. Experimental results show that the hybrid classifier performs better than a purely generative classifier (naive Bayes) or a purely discriminative classifier (Logistic Regression) a...

متن کامل

Two-Stage Text Classification Using Bayesian Networks

The“curse of dimensionality”provides a powerful impetus to explore alternative data structures and representations for text processing. This paper presents a method for preparing a dataset for classification by determining the utility of a very small number of related dimensions via a Discriminative Multinomial Naive Bayes process, then using these utility measurements to weight these dimension...

متن کامل

Discriminative Learning Using Boosted Generative Models

Discriminative learning, or learning for classification, is a common learning task that has been addressed in a variety of frameworks. One approach is to design a complex classifier, such as a support vector machine, that explicitly minimizes classification error. Alternatively, an ensemble of weak classifiers can be trained using boosting [4]. However, in some situations it may be desirable to...

متن کامل

Or gate Bayesian networks for text classification: A discriminative alternative approach to multinomial naive Bayes

We propose a simple Bayesian network-based text classifier, which may be considered as a discriminative counterpart of the generative multinomial naive Bayes classifier. The method relies on the use of a fixed network topology with the arcs going form term nodes to class nodes, and also on a network parametrization based on noisy or gates. Comparative experiments of the proposed method with nai...

متن کامل

On Discriminative Parameter Learning of Bayesian Network Classifiers

We introduce three discriminative parameter learning algorithms for Bayesian network classifiers based on optimizing either the conditional likelihood (CL) or a lower-bound surrogate of the CL. One training procedure is based on the extended Baum-Welch (EBW) algorithm. Similarly, the remaining two approaches iteratively optimize the parameters (initialized to ML) with a 2-step algorithm. In the...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2005